# Continuous pre-training
Consilience 40b 7Y9v38s5
Nous Consilience 40B is a 40-billion-parameter generative text model, pre-trained from scratch in a decentralized manner, supporting multiple languages and representing diverse human creative output.
Large Language Model Supports Multiple Languages
C
PsycheFoundation
44
3
Llama 3.1 Swallow 70B Instruct V0.3
Llama 3.1 Swallow is a series of large language models built on Meta Llama 3.1. It enhances Japanese language capabilities through continuous pre-training while retaining English language capabilities.
Large Language Model
Transformers Supports Multiple Languages

L
tokyotech-llm
1,659
12
Llama 3.1 Swallow 8B Instruct V0.3
Llama 3.1 Swallow is a series of large language models built on Meta Llama 3.1. It enhances Japanese capabilities through continuous pre-training while retaining English capabilities.
Large Language Model
Transformers Supports Multiple Languages

L
tokyotech-llm
16.48k
20
Llama SEA LION V3 8B
Llama-SEA-LION-v3-8B is a multilingual large language model optimized for Southeast Asian languages, supporting 11 Southeast Asian languages and having undergone continuous pre-training on approximately 200 billion tokens.
Large Language Model
Transformers Supports Multiple Languages

L
aisingapore
1,964
2
Mistral Nemo Japanese Instruct 2408
Apache-2.0
This is a Japanese continuous pre-trained model based on Mistral-Nemo-Instruct-2407, focusing on Japanese text generation tasks.
Large Language Model
Safetensors Supports Multiple Languages
M
cyberagent
1,898
39
Llama 3 Swallow 8B Instruct V0.1
A Japanese-optimized large language model built on Meta Llama 3, enhancing Japanese capabilities through continuous pre-training and improving instruction-following abilities through supervised fine-tuning.
Large Language Model
Transformers Supports Multiple Languages

L
tokyotech-llm
13.88k
20
Featured Recommended AI Models